Home
AI Tools
AI Models
MCP
AI NEWS
EN
Model Selection
Tags
Multi-GPU Inference
# Multi-GPU Inference
Deepseek R1 Distill Llama 70B FP8 Dynamic
MIT
The FP8 quantized version of DeepSeek-R1-Distill-Llama-70B, which optimizes inference performance by reducing the number of bits of weights and activations.
Large Language Model
Transformers
D
RedHatAI
45.77k
9
Featured Recommended AI Models
Empowering the Future, Your AI Solution Knowledge Base
English
简体中文
繁體中文
にほんご
© 2025
AIbase